Skip to content

[SYCL][PI][CUDA] Implements select kernel and kernel-group queries #1180

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 4 commits into from
Mar 4, 2020

Conversation

steffenlarsen
Copy link
Contributor

Implements the PI info query for PI_KERNEL_INFO_ATTRIBUTES, PI_KERNEL_PREFERRED_WORK_GROUP_SIZE_MULTIPLE, and PI_KERNEL_PRIVATE_MEM_SIZE.

@bader bader added the cuda CUDA back-end label Feb 25, 2020
@bader bader requested a review from smaslov-intel February 25, 2020 15:32
Copy link
Contributor

@romanovvlad romanovvlad left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Could you please add a test?

@steffenlarsen
Copy link
Contributor Author

Could you please add a test?

I've added a test. Host seems to miss implementations for some of these info queries so it has been excluded.

@steffenlarsen steffenlarsen force-pushed the steffen/cuda-kernel-info-queries branch from 1f34d5a to a6092f5 Compare March 4, 2020 16:27
@bader bader merged commit 5d72e6b into intel:sycl Mar 4, 2020
alexbatashev pushed a commit to alexbatashev/llvm that referenced this pull request Mar 5, 2020
…_accessor_refactor

* origin/sycl: (38 commits)
  [SYCL] Fix device::get_devices() with a non-host device type (intel#1235)
  [SYCL][PI][CUDA] Implement kernel and kernel-group information queries (intel#1180)
  [SYCL] Remove default error code value in exception (intel#1150)
  [SYCL] Fix devicelib assert LIT test (intel#1245)
  [SYCL] Set aux-target-cpu for SYCL offload device compilation (intel#1225)
  [SYCL] Remove fabs and ceil from the list of unsupported math functions (intel#1217)
  [SYCL] Fix circular reference between events and queues (intel#1226)
  [CI][Doc] Use SSH to deploy GitHub Pages (intel#1232)
  [SYCL][CUDA][Test] Testing for use of CUDA primary context (intel#1174)
  [SYCL] allow underscore symbol in temporary directory name
  [SYCL] Reject zero length arrays (intel#1153)
  [SYCL] Fix static code analyzis concerns (intel#1189)
  [SYCL] Add more details about the -fintelfpga option (intel#1218)
  [SYCL][CUDA] Select only NVPTX64 device binaries (intel#1223)
  [SYCL] Reverse max work-group size order (intel#1177)
  [SYCL][Doc] Add GroupAlgorithms extension (intel#1079)
  [SYCL] Fix SYCL internal enumerators conflict with user defined macro (intel#1188)
  [SYCL][CUDA] Fixes context release and unnamed context scope (intel#1207)
  [SYCL][CUDA] Fix context creation property parsing
  [CUDA][PI] clang-format pi.h
  ...
alexbatashev pushed a commit to alexbatashev/llvm that referenced this pull request Mar 5, 2020
…ctor_tests

* origin/sycl:
  [SYCL] Fix device::get_devices() with a non-host device type (intel#1235)
  [SYCL][PI][CUDA] Implement kernel and kernel-group information queries (intel#1180)
  [SYCL] Remove default error code value in exception (intel#1150)
  [SYCL] Fix devicelib assert LIT test (intel#1245)
  [SYCL] Set aux-target-cpu for SYCL offload device compilation (intel#1225)
  [SYCL] Remove fabs and ceil from the list of unsupported math functions (intel#1217)
fwyzard added a commit to cms-patatrack/patatrack-website that referenced this pull request Mar 5, 2020
vmaksimo pushed a commit to vmaksimo/llvm that referenced this pull request Sep 1, 2021
The BuiltIn variable/call name in SPV-IR should stick to
"__spirv_BuiltIn*", no matter what the SPIR-V linkage name is.

This is a regression of PR intel#1133.

Signed-off-by: Yilong Guo <[email protected]>

Original commit:
KhronosGroup/SPIRV-LLVM-Translator@894f95b
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
cuda CUDA back-end
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants